Goto

Collaborating Authors

 dpp model


MS-DPPs: Multi-Source Determinantal Point Processes for Contextual Diversity Refinement of Composite Attributes in Text to Image Retrieval

Sogi, Naoya, Shibata, Takashi, Terao, Makoto, Suganuma, Masanori, Okatani, Takayuki

arXiv.org Artificial Intelligence

Result diversification (RD) is a crucial technique in Text-to-Image Retrieval for enhancing the efficiency of a practical application. Conventional methods focus solely on increasing the diversity metric of image appearances. However, the diversity metric and its desired value vary depending on the application, which limits the applications of RD. This paper proposes a novel task called CDR-CA (Contextual Diversity Refinement of Composite Attributes). CDR-CA aims to refine the diversities of multiple attributes, according to the application's context. To address this task, we propose Multi-Source DPPs, a simple yet strong baseline that extends the Determinantal Point Process (DPP) to multi-sources. We model MS-DPP as a single DPP model with a unified similarity matrix based on a manifold representation. We also introduce Tangent Normalization to reflect contexts. Extensive experiments demonstrate the effectiveness of the proposed method.


A simple estimator of the correlation kernel matrix of a determinantal point process

Gouriéroux, Christian, Lu, Yang

arXiv.org Machine Learning

Determinantal Point Process (DPP) is a flexible family of distributions for random sets defined on the finite state space { 1, ...,d }, or equivalently for multivariate binary variables. This family is parameterized by either the L-ensemble kernel Σ, which is symmetric positive definite (SPD), or the correlation kernel matrix K, which is SPD, with eigenvalues lying strictly between 0 and 1. The literature has considered the maximum likelihood estimation (MLE) of Σ and K or its algorithmic analogues (Affandi et al., 2014; Brunel et al., 2017a,b), but it has since been shown that i) the likelihood function has at least 2


Duality induced by an embedding structure of determinantal point process

Hino, Hideitsu, Yano, Keisuke

arXiv.org Machine Learning

Specifically, we clarify the embedding structure of a DPP model in the exponential family of log-linear models (c.f., Agresti, 1990; Amari, 2001) in Theorem 1. Models embedded in exponential families are called curved exponential families. Information geometry (Amari, 1985) provides a measure, the e-embedding curvature tensor (Efron, 1975; Reeds, 1975; Amari, 1982; Sei, 2011), to quantify the extent to which a curved exponential family deviates from an exponential family. To check the e-embedding curvature as well as the Fisher information matrix, we apply the diagonal scaling (Marshall and Olkin, 1968), also known as the quality vs. diversity decomposition in the DPP literature (Kulesza and Taskar, 2012), to an L-ensemble kernel of a DPP model and then evaluate them, which clarifies that the subset of parameters related to the item-wise effects (quality terms) has zero e-embedding curvature (Corollary 1).


DPP-Based Adversarial Prompt Searching for Lanugage Models

Zhang, Xu, Wan, Xiaojun

arXiv.org Artificial Intelligence

Language models risk generating mindless and offensive content, which hinders their safe deployment. Therefore, it is crucial to discover and modify potential toxic outputs of pre-trained language models before deployment. In this work, we elicit toxic content by automatically searching for a prompt that directs pre-trained language models towards the generation of a specific target output. The problem is challenging due to the discrete nature of textual data and the considerable computational resources required for a single forward pass of the language model. To combat these challenges, we introduce Auto-regressive Selective Replacement Ascent (ASRA), a discrete optimization algorithm that selects prompts based on both quality and similarity with determinantal point process (DPP). Experimental results on six different pre-trained language models demonstrate the efficacy of ASRA for eliciting toxic content. Furthermore, our analysis reveals a strong correlation between the success rate of ASRA attacks and the perplexity of target outputs, while indicating limited association with the quantity of model parameters.


Deep Determinantal Point Processes

Gartrell, Mike, Dohmatob, Elvis

arXiv.org Machine Learning

Determinantal point processes (DPPs) have attracted significant attention as an elegant model that is able to capture the balance between quality and diversity within sets. DPPs are parameterized by a positive semi-definite kernel matrix. While DPPs have substantial expressive power, they are fundamentally limited by the parameterization of the kernel matrix and their inability to capture nonlinear interactions between items within sets. We present the deep DPP model as way to address these limitations, by using a deep feed-forward neural network to learn the kernel matrix. In addition to allowing us to capture nonlinear item interactions, the deep DPP also allows easy incorporation of item metadata into DPP learning. We show experimentally that the deep DPP can provide a considerable improvement in the predictive performance of DPPs.


Basket Completion with Multi-task Determinantal Point Processes

Warlop, Romain, Mary, Jérémie, Gartrell, Mike

arXiv.org Machine Learning

Determinantal point processes (DPPs) have received significant attention in the recent years as an elegant model for a variety of machine learning tasks, due to their ability to elegantly model set diversity and item quality or popularity. Recent work has shown that DPPs can be effective models for product recommendation and basket completion tasks. We present an enhanced DPP model that is specialized for the task of basket completion, the multi-task DPP. We view the basket completion problem as a multi-class classification problem, and leverage ideas from tensor factorization and multi-class classification to design the multi-task DPP model. We evaluate our model on several real-world datasets, and find that the multi-task DPP provides significantly better predictive quality than a number of state-of-the-art models.


Low-Rank Factorization of Determinantal Point Processes

Gartrell, Mike (Microsoft) | Paquet, Ulrich (Microsoft) | Koenigstein, Noam (Microsoft)

AAAI Conferences

Determinantal point processes (DPPs) have garnered attention as an elegant probabilistic model of set diversity. They are useful for a number of subset selection tasks, including product recommendation. DPPs are parametrized by a positive semi-definite kernel matrix. In this work we present a new method for learning the DPP kernel from observed data using a low-rank factorization of this kernel. We show that this low-rank factorization enables a learning algorithm that is nearly an order of magnitude faster than previous approaches, while also providing for a method for computing product recommendation predictions that is far faster (up to 20x faster or more for large item catalogs) than previous techniques that involve a full-rank DPP kernel. Furthermore, we show that our method provides equivalent or sometimes better test log-likelihood than prior full-rank DPP approaches.


The Bayesian Low-Rank Determinantal Point Process Mixture Model

Gartrell, Mike, Paquet, Ulrich, Koenigstein, Noam

arXiv.org Machine Learning

Determinantal point processes (DPPs) are an elegant model for encoding probabilities over subsets, such as shopping baskets, of a ground set, such as an item catalog. They are useful for a number of machine learning tasks, including product recommendation. DPPs are parametrized by a positive semi-definite kernel matrix. Recent work has shown that using a low-rank factorization of this kernel provides remarkable scalability improvements that open the door to training on large-scale datasets and computing online recommendations, both of which are infeasible with standard DPP models that use a full-rank kernel. In this paper we present a low-rank DPP mixture model that allows us to represent the latent structure present in observed subsets as a mixture of a number of component low-rank DPPs, where each component DPP is responsible for representing a portion of the observed data. The mixture model allows us to effectively address the capacity constraints of the low-rank DPP model. We present an efficient and scalable Markov Chain Monte Carlo (MCMC) learning algorithm for our model that uses Gibbs sampling and stochastic gradient Hamiltonian Monte Carlo (SGHMC). Using an evaluation on several real-world product recommendation datasets, we show that our low-rank DPP mixture model provides substantially better predictive performance than is possible with a single low-rank or full-rank DPP, and significantly better performance than several other competing recommendation methods in many cases.


Tweet Timeline Generation with Determinantal Point Processes

Yao, Jin-ge (Peking University) | Fan, Feifan (Peking University) | Zhao, Wayne Xin (Renmin University of China) | Wan, Xiaojun (Peking University) | Chang, Edward (HTC Research) | Xiao, Jianguo (Peking University)

AAAI Conferences

The task of tweet timeline generation (TTG) aims at selecting a small set of representative tweets to generate a meaningful timeline and providing enough coverage for a given topical query. This paper presents an approach based on determinantal point processes (DPPs) by jointly modeling the topical relevance of each selected tweet and overall selectional diversity. Aiming at better treatment for balancing relevance and diversity, we introduce two novel strategies, namely spectral rescaling and topical prior. Extensive experiments on the public TREC 2014 dataset demonstrate that our proposed DPP model along with the two strategies can achieve fairly competitive results against the state-of-the-art TTG systems.


Low-Rank Factorization of Determinantal Point Processes for Recommendation

Gartrell, Mike, Paquet, Ulrich, Koenigstein, Noam

arXiv.org Machine Learning

Determinantal point processes (DPPs) have garnered attention as an elegant probabilistic model of set diversity. They are useful for a number of subset selection tasks, including product recommendation. DPPs are parametrized by a positive semi-definite kernel matrix. In this work we present a new method for learning the DPP kernel from observed data using a low-rank factorization of this kernel. We show that this low-rank factorization enables a learning algorithm that is nearly an order of magnitude faster than previous approaches, while also providing for a method for computing product recommendation predictions that is far faster (up to 20x faster or more for large item catalogs) than previous techniques that involve a full-rank DPP kernel. Furthermore, we show that our method provides equivalent or sometimes better predictive performance than prior full-rank DPP approaches, and better performance than several other competing recommendation methods in many cases. We conduct an extensive experimental evaluation using several real-world datasets in the domain of product recommendation to demonstrate the utility of our method, along with its limitations.